Goto

Collaborating Authors

 energy cost


ELANA: A Simple Energy and Latency Analyzer for LLMs

Chiang, Hung-Yueh, Wang, Bokun, Marculescu, Diana

arXiv.org Artificial Intelligence

The latency and power consumption of large language models (LLMs) are major constraints when serving them across a wide spectrum of hardware platforms, from mobile edge devices to cloud GPU clusters. Benchmarking is crucial for optimizing efficiency in both model deployment and next-generation model development. To address this need, we open-source a simple profiling tool, \textbf{ELANA}, for evaluating LLMs. ELANA is designed as a lightweight, academic-friendly profiler for analyzing model size, key-value (KV) cache size, prefilling latency (Time-to-first-token, TTFT), generation latency (Time-per-output-token, TPOT), and end-to-end latency (Time-to-last-token, TTLT) of LLMs on both multi-GPU and edge GPU platforms. It supports all publicly available models on Hugging Face and offers a simple command-line interface, along with optional energy consumption logging. Moreover, ELANA is fully compatible with popular Hugging Face APIs and can be easily customized or adapted to compressed or low bit-width models, making it ideal for research on efficient LLMs or for small-scale proof-of-concept studies. We release the ELANA profiling tool at: https://github.com/enyac-group/Elana.


A Fast Heuristic Search Approach for Energy-Optimal Profile Routing for Electric Vehicles

Ahmadi, Saman, Jalili, Mahdi

arXiv.org Artificial Intelligence

We study the energy-optimal shortest path problem for electric vehicles (EVs) in large-scale road networks, where recuperated energy along downhill segments introduces negative energy costs. While traditional point-to-point pathfinding algorithms for EVs assume a known initial energy level, many real-world scenarios involving uncertainty in available energy require planning optimal paths for all possible initial energy levels, a task known as energy-optimal profile search. Existing solutions typically rely on specialized profile-merging procedures within a label-correcting framework that results in searching over complex profiles. In this paper, we propose a simple yet effective label-setting approach based on multi-objective A* search, which employs a novel profile dominance rule to avoid generating and handling complex profiles. We develop four variants of our method and evaluate them on real-world road networks enriched with realistic energy consumption data. Experimental results demonstrate that our energy profile A* search achieves performance comparable to energy-optimal A* with a known initial energy level.


Energy Costs and Neural Complexity Evolution in Changing Environments

Heesom-Green, Sian, Shock, Jonathan, Nitschke, Geoff

arXiv.org Artificial Intelligence

The Cognitive Buffer Hypothesis (CBH) posits that larger brains evolved to enhance survival in changing conditions. However, larger brains also carry higher energy demands, imposing additional metabolic burdens. Alongside brain size, brain organization plays a key role in cognitive ability and, with suitable architectures, may help mitigate energy challenges. This study evolves Artificial Neural Networks (ANNs) used by Reinforcement Learning (RL) agents to investigate how environmental variability and energy costs influence the evolution of neural complexity, defined in terms of ANN size and structure. Results indicate that under energy constraints, increasing seasonality led to smaller ANNs. This challenges CBH and supports the Expensive Brain Hypothesis (EBH), as highly seasonal environments reduced net energy intake and thereby constrained brain size. ANN structural complexity primarily emerged as a byproduct of size, where energy costs promoted the evolution of more efficient networks.



A Appendix

Neural Information Processing Systems

A.1 Proof of Proposition 3.2 First, we consider the solution of Eq. (9) for u( x) = kx . Thus, from the property of the martingale and It ˆ o's isometry formula, it follows that Eη (t) = Eη (0) = 0, Eη (t) So, to satisfy Condition (iii) in Theorem 2.2, we have to set Therefore, the exponential stability of the zero solution is assured. Now, applying Gronwall's inequality, we get E[ x (t) This therefore completes the proof of the whole theorem. A.3.2 Proof of Theorem 4.2 First we prove the estimation for E[ τ Applying It ˆ o's formula to log V (x) yields: log V ( x( t)) = log V (x Then, similar to the procedure for the energy cost in A.3.1, we can get that E[ x (t) Here we explain this term in more detail. The training for ES framework is not as efficient as AS.




An Empirical Study of Adder Neural Networks for Object Detection Xinghao Chen

Neural Information Processing Systems

Comparisons with state-of-the-arts are conducted on COCO and P ASCAL VOC benchmarks. Specifically, the proposed Adder FCOS achieves a 37.8% AP on the COCO val set, demonstrating comparable performance